An Access-mode Prediction Technique Based on Cache Hit and Miss Speculation for Cache Design Achieves Minimal Energy
نویسندگان
چکیده
The successful pursuit of high performance on computer systems has produced the negative by-product of high power dissipation. Circuit-level techniques alone can no longer keep power dissipation under a reasonable level. Researchers have made efforts to reduce power dissipation at the architectural level by producing such schemes as reducing on-chip cache power consumption—a major power consumer in microprocessors. For example, experts believe that the processor power consumed by on-chip caches will increase from the 15 percent in Compaq’s Alpha 21264 to 26 percent in future Alpha 21464 processors. A set-associative cache is commonly used in modern computer systems for its ability to reduce cache conflict misses. However, a conventional set-associative cache implementation is not power-efficient. As Figure 1a shows, a conventional n-way set-associative cache probes all n blocks (both tag and data portions) in a set but, at most, will only really use one block. The percentage of wasted energy will increase as cache associativity n increases. High-associativity caches already exist in some commercial processors. For example, the Intel Pentium 4 processor exploits four-way L1 caches and an eight-way L2 cache. An effective approach to reduce power consumption in set associative caches is lowering the number of memory cells involved in an access. One method divides each data RAM into multiple sub-banks and only activates words at the required offset from all cache ways. Another alternative is to selectively disable a subset of cache ways during execution periods with modest cache activity, or to dynamically resize both the number of sets and ways in caches. The phased cache first compares all the tags with the accessing address, then probes only the desired data way. Way-prediction is another effective approach that speculatively selects a way to access before making a normal cache access. Figures 1b and 1c illustrate the access patterns for phased and way-prediction n-way setassociative caches. Compared with the conventional implementation, the phased cache only probes one data subarray instead of n data subarrays (each way comprises a tag subarray and a data subarray). However, the sequential accesses of tag and data will increase the cache access latency. The way-prediction cache first accesses the tag and data subarrays of the predicted way. If the prediction is not correct, it then probes the rest of tag and data subarrays simultaneously. An access in a phased cache consumes more energy and has longer latency than a correctly predicted access in way-prediction cache, but consumes less energy than a mispredicted access. Hence, when the prediction accuracy is high, the way-prediction cache is more energy-efficient than the phased cache. Zhichun Zhu Xiaodong Zhang
منابع مشابه
Incorporating Pattern Prediction Technique for Energy Efficient Filter Cache Design
A filter cache is proposed at a higher level than the L1 (main) cache in the memory hierarchy and is much smaller. The typical size of filter cache is of the order of 512 Bytes. Prediction algorithms popularly based upon the Next Fetch Prediction Table (NFPT) helps making the choice between the filter cache and the main cache. In this paper we introduce a new prediction mechanism for predicting...
متن کاملStatic Pattern Predictor (SPP) Based Low Power Instruction Cache Design
Energy dissipation in cache memories is becoming a major design issue in embedded microprocessors. Predictive filter cache based instruction cache hierarchy is effective in reducing the access energy substantially at the cost of certain performance degradation. Here, the energy-delay product reduction heavily depends on the prediction accuracy of the predictor. In this paper, a simplified patte...
متن کاملCache Compression through Noise Prediction
Caches are very inefficiently utilized because not all the excess data fetched into the cache, to exploit spatial locality, is utilized. Our experiments show that Level 1 data cache has a utilization (in terms of the percentage of data brought into the cache that is actually used) of only about 45%. Our studies also show that a prediction accuracy of about 95% can be achieved when predicting th...
متن کاملEnergy-delay efficient filter cache hierarchy using pattern prediction scheme - Computers and Digital Techniques, IEE Proceedings-
Filter cache (FC) is an auxiliary cache much smaller than the main cache. The FC is closest in hierarchy to the instruction fetch unit and it must be small in size to achieve energyefficient realisations. A pattern prediction scheme is adapted to maximise energy savings in the FC hierarchy. The pattern prediction mechanism proposed relies on the spatial hit or miss pattern of the instruction ac...
متن کاملEnhancing Cache Performance Based on Improved Average Access Time
A high performance computer includes a fast processor and millions bytes of memory. During the data processing, huge amount of information are shuffled between the memory and processor. Because of its small size and its effectiveness speed, cache has become a common feature of high performance computers. Enhancing cache performance proved to be essential in the speed up of cache-based computers...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
عنوان ژورنال:
دوره شماره
صفحات -
تاریخ انتشار 2002